1,682 research outputs found

    Interior-point solver for convex separable block-angular problems

    Get PDF
    Constraints matrices with block-angular structures are pervasive in Optimization. Interior-point methods have shown to be competitive for these structured problems by exploiting the linear algebra. One of these approaches solved the normal equations using sparse Cholesky factorizations for the block constraints, and a preconditioned conjugate gradient (PCG) for the linking constraints. The preconditioner is based on a power series expansion which approximates the inverse of the matrix of the linking constraints system. In this work we present an efficient solver based on this algorithm. Some of its features are: it solves linearly constrained convex separable problems (linear, quadratic or nonlinear); both Newton and second-order predictor-corrector directions can be used, either with the Cholesky+PCG scheme or with a Cholesky factorization of normal equations; the preconditioner may include any number of terms of the power series; for any number of these terms, it estimates the spectral radius of the matrix in the power series (which is instrumental for the quality of the precondi- tioner). The solver has been hooked to SML, a structure-conveying modelling language based on the popular AMPL modeling language. Computational results are reported for some large and/or difficult instances in the literature: (1) multicommodity flow problems; (2) minimum congestion problems; (3) statistical data protection problems using l1 and l2 distances (which are linear and quadratic problems, respectively), and the pseudo-Huber function, a nonlinear approximation to l1 which improves the preconditioner. In the largest instances, of up to 25 millions of variables and 300000 constraints, this approach is from two to three orders of magnitude faster than state-of-the-art linear and quadratic optimization solvers.Preprin

    A specialized interior-point algorithm for huge minimum convex cost flows in bipartite networks

    Get PDF
    Research Report UPC-DEIO DR 2018-01. November 2018The computation of the Newton direction is the most time consuming step of interior-point methods. This direction was efficiently computed by a combination of Cholesky factorizations and conjugate gradients in a specialized interior-point method for block-angular structured problems. In this work we apply this algorithmic approach to solve very large instances of minimum cost flows problems in bipartite networks, for convex objective functions with diagonal Hessians (i.e., either linear, quadratic or separable nonlinear objectives). After analyzing the theoretical properties of the interior-point method for this kind of problems, we provide extensive computational experiments with linear and quadratic instances of up to one billion arcs and 200 and five million nodes in each subset of the node partition. For linear and quadratic instances our approach is compared with the barriers algorithms of CPLEX (both standard path-following and homogeneous-self-dual); for linear instances it is also compared with the different algorithms of the state-of-the-art network flow solver LEMON (namely: network simplex, capacity scaling, cost scaling and cycle canceling). The specialized interior-point approach significantly outperformed the other approaches in most of the linear and quadratic transportation instances tested. In particular, it always provided a solution within the time limit and it never exhausted the 192 Gigabytes of memory of the server used for the runs. For assignment problems the network algorithms in LEMON were the most efficient option.Peer ReviewedPreprin

    Revisiting interval protection, a.k.a. partial cell suppression, for tabular data

    Get PDF
    The final publication is available at link.springer.comInterval protection or partial cell suppression was introduced in “M. Fischetti, J.-J. Salazar, Partial cell suppression: A new methodology for statistical disclosure control, Statistics and Computing, 13, 13–21, 2003” as a “linearization” of the difficult cell suppression problem. Interval protection replaces some cells by intervals containing the original cell value, unlike in cell suppression where the values are suppressed. Although the resulting optimization problem is still huge—as in cell suppression, it is linear, thus allowing the application of efficient procedures. In this work we present preliminary results with a prototype implementation of Benders decomposition for interval protection. Although the above seminal publication about partial cell suppression applied a similar methodology, our approach differs in two aspects: (i) the boundaries of the intervals are completely independent in our implementation, whereas the one of 2003 solved a simpler variant where boundaries must satisfy a certain ratio; (ii) our prototype is applied to a set of seven general and hierarchical tables, whereas only three two-dimensional tables were solved with the implementation of 2003.Peer ReviewedPostprint (author's final draft

    A linear optimization based method for data privacy in statistical tabular data

    Get PDF
    National Statistical Agencies routinely disseminate large amounts of data. Prior to dissemination these data have to be protected to avoid releasing confidential information. Controlled tabular adjustment (CTA) is one of the available methods for this purpose. CTA formulates an optimization problem that looks for the safe table which is closest to the original one. The standard CTA approach results in a mixed integer linear optimization (MILO) problem, which is very challenging for current technology. In this work we present a much less costly variant of CTA that formulates a multiobjective linear optimization (LO) problem, where binary variables are pre-fixed, and the resulting continuous problem is solved by lexicographic optimization. Extensive computational results are reported using both commercial (CPLEX and XPRESS) and open source (Clp) solvers, with either simplex or interior-point methods, on a set of real instances. Most instances were successfully solved with the LO-CTA variant in less than one hour, while many of them are computationally very expensive with the MILO-CTA formulation. The interior-point method outperformed simplex in this particular application.Peer ReviewedPreprin

    Implementación de un algoritmo primal-dual de orden superior mediante el uso de un método predictor-corrector para programación lineal

    Get PDF
    Se presenta una implementación de un algoritmo primal-dual de punto interior para la solución de problemas lineales. El algoritmo difiere de otros ya existentes (como el implementado en el sistema LoQo) en el hecho de que soluciona las denominadas "ecuaciones normales en forma primal" (LoQo soluciona el denominado "sistema aumentado") y en que realiza una clara distinción entre variables acotadas superior e inferiormente, y aquéllas sólo acotadas inferiormente. La eficiencia de la implementación es comparada con el sistema LoQo. Para la comparación se utilizan 80 problemas lineales de la colección Netlib (Gay, 1985), una batería estándar de problemas de programación lineal. Este trabajo es el primero de una serie de dos, cuyo objetivo es la resolución eficiente de problemas cuadráticos por técnicas de punto interior

    Stabilized Benders methods for large-scale combinatorial optimization, with appllication to data privacy

    Get PDF
    The Cell Suppression Problem (CSP) is a challenging Mixed-Integer Linear Problem arising in statistical tabular data protection. Medium sized instances of CSP involve thousands of binary variables and million of continuous variables and constraints. However, CSP has the typical structure that allows application of the renowned Benders’ decomposition method: once the “complicating” binary variables are fixed, the problem decomposes into a large set of linear subproblems on the “easy” continuous ones. This allows to project away the easy variables, reducing to a master problem in the complicating ones where the value functions of the subproblems are approximated with the standard cutting-plane approach. Hence, Benders’ decomposition suffers from the same drawbacks of the cutting-plane method, i.e., oscillation and slow convergence, compounded with the fact that the master problem is combinatorial. To overcome this drawback we present a stabilized Benders decomposition whose master is restricted to a neighborhood of successful candidates by local branching constraints, which are dynamically adjusted, and even dropped, during the iterations. Our experiments with randomly generated and real-world CSP instances with up to 3600 binary variables, 90M continuous variables and 15M inequality constraints show that our approach is competitive with both the current state-of-the-art (cutting-plane-based) code for cell suppression, and the Benders implementation in CPLEX 12.7. In some instances, stabilized Benders is able to quickly provide a very good solution in less than one minute, while the other approaches were not able to find any feasible solution in one hour.Peer ReviewedPreprin

    Improving an interior-point algorithm for multicommodity flows by quadratic regularizations

    Get PDF
    One of the best approaches for some classes of multicommodity flow problems is a specialized interior-point method that solves the normal equations by a combination of Cholesky factorizations and preconditioned conjugate gradient. Its efficiency depends on the spectral radius—in [0,1)—of a certain matrix in the definition of the preconditioner. In a recent work the authors improved this algorithm (i.e., reduced the spectral radius) for general block-angular problems by adding a quadratic regularization to the logarithmic barrier. This barrier was shown to be self-concordant, which guarantees the convergence and polynomial complexity of the algorithm. In this work we focus on linear multicommodity problems, a particular case of primal block-angular ones. General results are tailored for multicommodity flows, allowing a local sensitivity analysis on the effect of the regularization. Extensive computational results on some standard and some difficult instances, testing several regularization strategies, are also provided. These results show that the regularized interior-point algorithm is more efficient than the nonregularized one. From this work it can be concluded that, if interior-point methods based on conjugate gradients are used, linear multicommodity flow problems are most efficiently solved as a sequence of quadratic ones.Preprin

    On assessing the disclosure risk of controlled adjustment methods for statistical tabular data

    Get PDF
    Minimum distance controlled tabular adjustment is a recent perturbative approach for statistical disclosure control in tabular data. Given a table to be protected, it looks for the closest safe table, using some particular distance. Con trolled adjustment is known to provide high data utility. However, the disclosure risk has only been partially analyzed using theoretical results from optimization. This work ext ends these previous results, providing both a more detailed theoretical analysis, and an extensive empirical assess- ment of the disclosure risk of the method. A set of 25 instance s from the literature and four different attacker scenarios are considered, with sever al random replications for each scenario, both for L 1 and L 2 distances. This amounts to the solution of more than 2000 optimization problems. The analysis of the results shows th at the approach has low dis- closure risk when the attacker has no good information on the bounds of the optimization problem. On the other hand, when the attacker has good estima tes of the bounds, and the only uncertainty is in the objective function (which is a very strong assumption), the disclosure risk of controlled adjustment is high and it s hould be avoided.Peer ReviewedPreprin

    Assessing the disclosure risk of CTA-like methods

    Get PDF
    Minimum distance controlled tabular adjustment (CTA) is a recent perturbative approach for statistical disclosure control in tabular data. CTA looks for the closest safe table, using some particular distance. In this talk we provide empirical results to assess the disclosure risk of the method. A set of 33 instances from the literature and four different attacker scenarios are considered. The result s show that, unless the attacker has good information about the original table, CTA has low disclosure risk. This talk summarizes results reported in the paper “Castro, J. (2013). On assessing the disclosure risk of controlled adjustment methods for statistical tabular data, International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems, 20, 921–941.Peer ReviewedPostprint (published version

    Thirty years of optimization-based SDC methods for tabular data

    Get PDF
    In 1966 Bacharach published in Management Science a work on matrix rounding problems in two-way tables of economic statistics, formulated as a network optimization problem. This is likely the first application of optimization/operations research for statistical disclosure control (SDC) in tabular data. Years later, in 1982, Cox and Ernst used the same approach in a work in INFOR for a similar problem: controlled rounding. And thirty years ago, in 1992, a paper by Kelly, Golden and Assad appeared in Networks about the solution of the cell suppression problem, also using network optimization. Cell suppression was used for years as the main SDC technique for tabular data, and it was an active field of research which resulted in several lines of work and many publications. The above are some of the seminal works on the use of optimization methods for SDC when releasing tabular data. This paper discusses some of the research done this field since then, with a focus on the approaches that were of practical use. It also discusses their pros and cons compared to recent techniques that are not based on optimization methods.Peer ReviewedPostprint (published version
    corecore